Response-Time Bounds for Concurrent GPU Scheduling
نویسنده
چکیده
Graphics processing units (GPUs) have been receiving increasing attention in the real-time systems community as a potential solution for hosting workloads like those found in autonomous-driving use cases that require significant computational capacity. Allowing multiple programs to access a GPU concurrently can enable the GPU to be more efficiently utilized, if each individual program is incapable of occupying all GPU resources. In this work, we summarize the basic scheduling rules for concurrent GPU scheduling in NVIDIA GPUs. We define a task model for GPU scheduling based on these scheduling rules. In ongoing work, we are attempting to obtain response-time bounds for tasks under this model.
منابع مشابه
GPU Sharing for Image Processing in Embedded Real-Time Systems∗
To more efficiently utilize graphics processing units (GPUs) when supporting real-time workloads, it may be beneficial to allow multiple tasks to issue GPU computations without blocking one another. For such an option to be viable, it is necessary to know the extent to which concurrent GPU computations interfere with each other when accessing hardware resources. In this paper, measurement data ...
متن کاملInferring Scheduling Policies of an Embedded CUDA GPU∗
Embedded systems augmented with graphics processing units (GPUs) are seeing increased use in safety-critical real-time systems such as autonomous vehicles. Due to monetary cost requirements along with size, weight, and power (SWaP) constraints, embedded GPUs are often computationally impoverished compared to those used in non-embedded systems. In order to maximize performance on these impoveris...
متن کاملConcurrent Locomotive Assignment and Freight Train Scheduling
The locomotive assignment and the freight train scheduling are important problems in railway transportation. Freight cars are coupled to form a freight rake. The freight rake becomes a train when a locomotive is coupled to it. The locomotive assignment problem assigns locomotives to a set of freight rakes in a way that, with minimum locomotive deadheading time, rake coupling delay and locomotiv...
متن کاملImplementation of the direction of arrival estimation algorithms by means of GPU-parallel processing in the Kuda environment (Research Article)
Direction-of-arrival (DOA) estimation of audio signals is critical in different areas, including electronic war, sonar, etc. The beamforming methods like Minimum Variance Distortionless Response (MVDR), Delay-and-Sum (DAS), and subspace-based Multiple Signal Classification (MUSIC) are the most known DOA estimation techniques. The mentioned methods have high computational complexity. Hence using...
متن کاملConcurrent Analytical Query Processing with GPUs
In current databases, GPUs are used as dedicated accelerators to process each individual query. Sharing GPUs among concurrent queries is not supported, causing serious resource underutilization. Based on the profiling of an opensource GPU query engine running commonly used singlequery data warehousing workloads, we observe that the utilization of main GPU resources is only up to 25%. The underu...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017